Edge Detection
   HOME

TheInfoList



OR:

Edge detection includes a variety of
mathematical Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
methods that aim at identifying edges,
curve In mathematics, a curve (also called a curved line in older texts) is an object similar to a line (geometry), line, but that does not have to be Linearity, straight. Intuitively, a curve may be thought of as the trace left by a moving point (ge ...
s in a
digital image A digital image is an image composed of picture elements, also known as ''pixels'', each with ''finite'', '' discrete quantities'' of numeric representation for its intensity or gray level that is an output from its two-dimensional functions ...
at which the image brightness changes sharply or, more formally, has discontinuities. The same problem of finding discontinuities in one-dimensional signals is known as ''
step detection In statistics and signal processing, step detection (also known as step smoothing, step filtering, shift detection, jump detection or edge detection) is the process of finding abrupt changes (steps, jumps, shifts) in the mean level of a time seri ...
'' and the problem of finding signal discontinuities over time is known as ''
change detection In statistical analysis, change detection or change point detection tries to identify times when the probability distribution of a stochastic process or time series changes. In general the problem concerns both detecting whether or not a change ...
''. Edge detection is a fundamental tool in
image processing An image is a visual representation of something. It can be two-dimensional, three-dimensional, or somehow otherwise feed into the visual system to convey information. An image can be an artifact, such as a photograph or other two-dimensiona ...
,
machine vision Machine vision (MV) is the technology and methods used to provide imaging-based automatic inspection and analysis for such applications as automatic inspection, process control, and robot guidance, usually in industry. Machine vision refers to ...
and
computer vision Computer vision is an interdisciplinary scientific field that deals with how computers can gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to understand and automate tasks that the hum ...
, particularly in the areas of feature detection and
feature extraction In machine learning, pattern recognition, and image processing, feature extraction starts from an initial set of measured data and builds derived values (features) intended to be informative and non-redundant, facilitating the subsequent learning a ...
.


Motivations

The purpose of detecting sharp changes in image brightness is to capture important events and changes in properties of the world. It can be shown that under rather general assumptions for an image formation model, discontinuities in image brightness are likely to correspond to: * discontinuities in depth, * discontinuities in surface orientation, * changes in material properties and * variations in scene illumination. In the ideal case, the result of applying an edge detector to an image may lead to a set of connected curves that indicate the boundaries of objects, the boundaries of surface markings as well as curves that correspond to discontinuities in surface orientation. Thus, applying an edge detection algorithm to an image may significantly reduce the amount of data to be processed and may therefore filter out information that may be regarded as less relevant, while preserving the important structural properties of an image. If the edge detection step is successful, the subsequent task of interpreting the information contents in the original image may therefore be substantially simplified. However, it is not always possible to obtain such ideal edges from real life images of moderate complexity. Edges extracted from non-trivial images are often hampered by ''fragmentation'', meaning that the edge curves are not connected, missing edge segments as well as ''false edges'' not corresponding to interesting phenomena in the image – thus complicating the subsequent task of interpreting the image data.T. Lindeberg (1998) "Edge detection and ridge detection with automatic scale selection", International Journal of Computer Vision, 30, 2, pages 117–154.
/ref> Edge detection is one of the fundamental steps in image processing, image analysis, image pattern recognition, and computer vision techniques.


Edge properties

The edges extracted from a two-dimensional image of a three-dimensional scene can be classified as either viewpoint dependent or viewpoint independent. A ''viewpoint independent edge'' typically reflects inherent properties of the three-dimensional objects, such as surface markings and surface shape. A ''viewpoint dependent edge'' may change as the viewpoint changes, and typically reflects the geometry of the scene, such as objects occluding one another. A typical edge might for instance be the border between a block of red color and a block of yellow. In contrast a line (as can be extracted by a ridge detector) can be a small number of
pixel In digital imaging, a pixel (abbreviated px), pel, or picture element is the smallest addressable element in a raster image, or the smallest point in an all points addressable display device. In most digital display devices, pixels are the smal ...
s of a different color on an otherwise unchanging background. For a line, there may therefore usually be one edge on each side of the line.


A simple edge model

Although certain literature has considered the detection of ideal step edges, the edges obtained from natural images are usually not at all ideal step edges. Instead they are normally affected by one or several of the following effects: * focal blur caused by a finite
depth-of-field The depth of field (DOF) is the distance between the nearest and the furthest objects that are in acceptably sharp focus in an image captured with a camera. Factors affecting depth of field For cameras that can only focus on one object dist ...
and finite point spread function. * penumbral blur caused by shadows created by light sources of non-zero radius. *
shading Shading refers to the depiction of depth perception in 3D models (within the field of 3D computer graphics) or illustrations (in visual art) by varying the level of darkness. Shading tries to approximate local behavior of light on the object's ...
at a smooth object A number of researchers have used a Gaussian smoothed step edge (an
error function In mathematics, the error function (also called the Gauss error function), often denoted by , is a complex function of a complex variable defined as: :\operatorname z = \frac\int_0^z e^\,\mathrm dt. This integral is a special (non-elementary ...
) as the simplest extension of the ideal step edge model for modeling the effects of edge blur in practical applications. Thus, a one-dimensional image f that has exactly one edge placed at x = 0 may be modeled as: :f(x) = \frac \left( \operatorname\left(\frac\right) + 1\right) + I_\ell. At the left side of the edge, the intensity is I_\ell = \lim_ f(x), and right of the edge it is I_r = \lim_ f(x). The scale parameter \sigma is called the blur scale of the edge. Ideally this scale parameter should be adjusted based on the quality of image to avoid destroying true edges of the image.


Why it is a non-trivial task

To illustrate why edge detection is not a trivial task, consider the problem of detecting edges in the following one-dimensional signal. Here, we may intuitively say that there should be an edge between the 4th and 5th pixels. If the intensity difference were smaller between the 4th and the 5th pixels and if the intensity differences between the adjacent neighboring pixels were higher, it would not be as easy to say that there should be an edge in the corresponding region. Moreover, one could argue that this case is one in which there is more than one edge. Hence, to firmly state a specific threshold on how large the intensity change between two neighbouring pixels must be for us to say that there should be an edge between these pixels is not always simple. Indeed, this is one of the reasons why edge detection may be a non-trivial problem unless the objects in the scene are particularly simple and the illumination conditions can be well controlled (see for example, the edges extracted from the image with the girl above).


Approaches

There are many methods for edge detection, but most of them can be grouped into two categories, search-based and zero-crossing based. The search-based methods detect edges by first computing a measure of edge strength, usually a
first-order derivative expression In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). Derivatives are a fundamental tool of calculus. F ...
such as the gradient magnitude, and then searching for local directional maxima of the gradient magnitude using a computed estimate of the local orientation of the edge, usually the gradient direction. The zero-crossing based methods search for zero crossings in a
second-order derivative expression In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). Derivatives are a fundamental tool of calculus. ...
computed from the image in order to find edges, usually the zero-crossings of the
Laplacian In mathematics, the Laplace operator or Laplacian is a differential operator given by the divergence of the gradient of a scalar function on Euclidean space. It is usually denoted by the symbols \nabla\cdot\nabla, \nabla^2 (where \nabla is the ...
or the zero-crossings of a non-linear differential expression. As a pre-processing step to edge detection, a smoothing stage, typically
Gaussian smoothing In image processing, a Gaussian blur (also known as Gaussian smoothing) is the result of blurring an image by a Gaussian function (named after mathematician and scientist Carl Friedrich Gauss). It is a widely used effect in graphics software, ...
, is almost always applied (see also
noise reduction Noise reduction is the process of removing noise from a signal. Noise reduction techniques exist for audio and images. Noise reduction algorithms may distort the signal to some degree. Noise rejection is the ability of a circuit to isolate an und ...
). The edge detection methods that have been published mainly differ in the types of smoothing filters that are applied and the way the measures of edge strength are computed. As many edge detection methods rely on the computation of image gradients, they also differ in the types of filters used for computing gradient estimates in the ''x''- and ''y''-directions. A survey of a number of different edge detection methods can be found in (Ziou and Tabbone 1998); see also the encyclopedia articles on edge detection in ''Encyclopedia of Mathematics'' and Encyclopedia of Computer Science and Engineering.


Canny

John Canny John F. Canny (born in 1958) is an Australian computer scientist, and '' Paul E Jacobs and Stacy Jacobs Distinguished Professor of Engineering'' in the Computer Science Department of the University of California, Berkeley. He has made significant ...
considered the mathematical problem of deriving an optimal smoothing filter given the criteria of detection, localization and minimizing multiple responses to a single edge. He showed that the optimal filter given these assumptions is a sum of four exponential terms. He also showed that this filter can be well approximated by first-order derivatives of Gaussians. Canny also introduced the notion of non-maximum suppression, which means that given the presmoothing filters, edge points are defined as points where the gradient magnitude assumes a local maximum in the gradient direction. Looking for the zero crossing of the 2nd derivative along the gradient direction was first proposed by
Haralick Robert M. Haralick (born 1943) is Distinguished Professor in Computer Science at Graduate Center of the City University of New York (CUNY). Haralick is one of the leading figures in computer vision, pattern recognition, and image analysis. He is a ...
. It took less than two decades to find a modern geometric variational meaning for that operator that links it to the Marr–Hildreth (zero crossing of the Laplacian) edge detector. That observation was presented by
Ron Kimmel Ron Kimmel ( he, רון קימל, b. 1963) is a professor of Computer Science and Electrical and Computer Engineering (by courtesy) at the Technion Israel Institute of Technology. He holds a D.Sc. degree in electrical engineering (1995) from the ...
and Alfred Bruckstein. Although his work was done in the early days of computer vision, the Canny edge detector (including its variations) is still a state-of-the-art edge detector. Edge detectors that perform better than the Canny usually require longer computation times or a greater number of parameters.


Kovalevsky

Vladimir A. Kovalevsky has suggested a quite different approach. He uses a preprocessing of the image with the Sigma filter and with a special filter for the dilution of the ramps. This method uses no brightness of the image but only the intensities of the color channels which is important for detecting an edge between two adjacent pixels of equal brightness but different colors. The method scans the image two times: first along the horizontal lines and second along the vertical columns. In each horizontal line six consequent adjacent pixels are considered and five color difference between each two adjacent pixels are calculated. Each color difference is the sum of absolute differences of the intensities of the color channels Red, Green, and Blue of the corresponding adjacent pixels. If this sum is greater than a given threshold, then the sign of the color difference is set equal to the sign of the difference of the green intensities. If the green difference is zero, then the sign of the color difference is set equal to the sign of the difference of the red intensities. If, however, both the green and the red differences are zero, then the sign of the color difference is set equal to the sign of the blue difference which in this case cannot be zero since the sum is greater than the threshold. Certain conditions for the values and signs of the five color differences are specified in such way that if the conditions are fulfilled, then a short vertical stroke is put between the third and the fourth of the six pixels as the label of the edge. Similar calculations are performed for the vertical columns. In this case a short horizontal stroke is put between the third and the fourth of the six subsequent pixels. The vertical and horizontal strokes (being the one-dimensional cells of an abstract cell complex corresponding to the image) mostly compose a connected sequence representing the edge. This method is robust and very fast and, what is more important, it can detect edges between adjacent pixels of equal brightness’s if the color difference between these pixels is greater than the threshold. The Canny–Deriche detector was derived from similar mathematical criteria as the Canny edge detector, although starting from a discrete viewpoint and then leading to a set of recursive filters for image smoothing instead of exponential filters or Gaussian filters. The differential edge detector described below can be seen as a reformulation of Canny's method from the viewpoint of differential invariants computed from a
scale space representation Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal th ...
leading to a number of advantages in terms of both theoretical analysis and sub-pixel implementation. In that aspect,
Log Gabor filter In signal processing it is useful to simultaneously analyze the space and frequency characteristics of a signal. While the Fourier transform gives the frequency information of the signal, it is not localized. This means that we cannot determine whic ...
have been shown to be a good choice to extract boundaries in natural scenes.


Other first-order methods

Different gradient operators can be applied to estimate image gradients from the input image or a smoothed version of it. The simplest approach is to use central differences: : \begin L_x(x, y) & = -\frac 1 2 L(x-1, y) + 0 \cdot L(x, y) + \frac 1 2 \cdot L(x+1, y) \\ ptL_y(x, y) & = -\frac1 2 L(x, y-1) + 0 \cdot L(x, y) + \frac 1 2 \cdot L(x, y+1), \end corresponding to the application of the following filter masks to the image data: : L_y = \begin +1/2 & 0 & -1/2 \end L \quad \text \quad L_x = \begin +1/2 \\ 0 \\ -1/2 \end L. The well-known and earlier
Sobel operator The Sobel operator, sometimes called the Sobel–Feldman operator or Sobel filter, is used in image processing and computer vision, particularly within edge detection algorithms where it creates an image emphasising edges. It is named after ...
is based on the following filters: : L_y = \begin +1 & 0 & -1 \\ +2 & 0 & -2 \\ +1 & 0 & -1 \end L \quad \text \quad L_x = \begin +1 & +2 & +1 \\ 0 & 0 & 0 \\ -1 & -2 & -1 \end L. Given such estimates of first-order
image derivative Image derivatives can be computed by using small convolution filters of size 2 × 2 or 3 × 3, such as the Discrete Laplace operator, Laplacian, Sobel operator, Sobel, Roberts cross, Roberts and Prewitt operator, Prewitt operato ...
s, the gradient magnitude is then computed as: :, \nabla L, = \sqrt while the gradient orientation can be estimated as :\theta = \operatorname(L_y, L_x). Other first-order difference operators for estimating image gradient have been proposed in the
Prewitt operator The Prewitt operator is used in image processing, particularly within edge detection algorithms. Technically, it is a discrete differentiation operator, computing an approximation of the gradient of the image intensity function. At each point i ...
,
Roberts cross The Roberts cross operator is used in image processing and computer vision for edge detection. It was one of the first edge detectors and was initially proposed by Lawrence Roberts in 1963. As a differential operator, the idea behind the Rob ...
, Kayyali operator and Frei–Chen operator. It is possible to extend filters dimension to avoid the issue of recognizing edge in low SNR image. The cost of this operation is loss in terms of resolution. Examples are Extended Prewitt 7×7.


Thresholding and linking

Once we have computed a measure of edge strength (typically the gradient magnitude), the next stage is to apply a threshold, to decide whether edges are present or not at an image point. The lower the threshold, the more edges will be detected, and the result will be increasingly susceptible to
noise Noise is unwanted sound considered unpleasant, loud or disruptive to hearing. From a physics standpoint, there is no distinction between noise and desired sound, as both are vibrations through a medium, such as air or water. The difference arise ...
and detecting edges of irrelevant features in the image. Conversely a high threshold may miss subtle edges, or result in fragmented edges. If the edge is applied to just the gradient magnitude image, the resulting edges will in general be thick and some type of edge thinning post-processing is necessary. For edges detected with non-maximum suppression however, the edge curves are thin by definition and the edge pixels can be linked into edge polygon by an edge linking (edge tracking) procedure. On a discrete grid, the non-maximum suppression stage can be implemented by estimating the gradient direction using first-order derivatives, then rounding off the gradient direction to multiples of 45 degrees, and finally comparing the values of the gradient magnitude in the estimated gradient direction. A commonly used approach to handle the problem of appropriate thresholds for thresholding is by using thresholding with
hysteresis Hysteresis is the dependence of the state of a system on its history. For example, a magnet may have more than one possible magnetic moment in a given magnetic field, depending on how the field changed in the past. Plots of a single component of ...
. This method uses multiple thresholds to find edges. We begin by using the upper threshold to find the start of an edge. Once we have a start point, we then trace the path of the edge through the image pixel by pixel, marking an edge whenever we are above the lower threshold. We stop marking our edge only when the value falls below our lower threshold. This approach makes the assumption that edges are likely to be in continuous curves, and allows us to follow a faint section of an edge we have previously seen, without meaning that every noisy pixel in the image is marked down as an edge. Still, however, we have the problem of choosing appropriate thresholding parameters, and suitable thresholding values may vary over the image.


Edge thinning

Edge thinning is a technique used to remove the unwanted spurious points on the edges in an image. This technique is employed after the image has been filtered for noise (using median, Gaussian filter etc.), the edge operator has been applied (like the ones described above, Canny or Sobel) to detect the edges and after the edges have been smoothed using an appropriate threshold value. This removes all the unwanted points and if applied carefully, results in one pixel thick edge elements. Advantages: # Sharp and thin edges lead to greater efficiency in
object recognition Object recognition – technology in the field of computer vision for finding and identifying objects in an image or video sequence. Humans recognize a multitude of objects in images with little effort, despite the fact that the image of the ...
. # If Hough transforms are used to detect lines and ellipses, then thinning could give much better results. # If the edge happens to be the boundary of a region, then thinning could easily give the image parameters like perimeter without much algebra. There are many popular algorithms used to do this, one such is described below: # Choose a type of
connectivity Connectivity may refer to: Computing and technology * Connectivity (media), the ability of the social media to accumulate economic capital from the users connections and activities * Internet connectivity, the means by which individual terminal ...
, like 8, 6 or 4. # 8 connectivity is preferred, where all the immediate pixels surrounding a particular pixel are considered. # Remove points from North, south, east and west. # Do this in multiple passes, i.e. after the north pass, use the same semi processed image in the other passes and so on. # Remove a point if:
The point has no neighbors in the North (if you are in the north pass, and respective directions for other passes).
The point is not the end of a line.
The point is isolated.
Removing the points will not cause to disconnect its neighbors in any way. # Else keep the point. The number of passes across direction should be chosen according to the level of accuracy desired.


Second-order approaches

Some edge-detection operators are instead based upon second-order derivatives of the intensity. This essentially captures the rate of change in the intensity gradient. Thus, in the ideal continuous case, detection of zero-crossings in the second derivative captures local maxima in the gradient. The early Marr–Hildreth operator is based on the detection of zero-crossings of the Laplacian operator applied to a Gaussian-smoothed image. It can be shown, however, that this operator will also return false edges corresponding to local minima of the gradient magnitude. Moreover, this operator will give poor localization at curved edges. Hence, this operator is today mainly of historical interest.


Differential

A more refined second-order edge detection approach which automatically detects edges with sub-pixel accuracy, uses the following ''differential approach'' of detecting zero-crossings of the second-order directional derivative in the gradient direction: Following the differential geometric way of expressing the requirement of non-maximum suppression proposed by Lindeberg,T. Lindeberg (1993) "Discrete derivative approximations with scale-space properties: A basis for low-level feature extraction", J. of Mathematical Imaging and Vision, 3(4), pages 349–376.
/ref> let us introduce at every image point a local coordinate system (u, v), with the v-direction parallel to the gradient direction. Assuming that the image has been pre-smoothed by Gaussian smoothing and a
scale space representation Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal th ...
L(x, y; t) at scale t has been computed, we can require that the gradient magnitude of the
scale space representation Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal th ...
, which is equal to the first-order directional derivative in the v-direction L_v, should have its first order directional derivative in the v-direction equal to zero :\partial_v(L_v) = 0 while the second-order directional derivative in the v-direction of L_v should be negative, i.e., :\partial_(L_v) \leq 0. Written out as an explicit expression in terms of local partial derivatives L_x, L_y, \ldots , L_, this edge definition can be expressed as the zero-crossing curves of the differential invariant :L_v^2 L_ = L_x^2 \, L_ + 2 \, L_x \, L_y \, L_ + L_y^2 \, L_ = 0, that satisfy a sign-condition on the following differential invariant :L_v^3 L_ = L_x^3 \, L_ + 3 \, L_x^2 \, L_y \, L_ + 3 \, L_x \, L_y^2 \, L_ + L_y^3 \, L_ \leq 0 where L_x, L_y, \ldots , L_ denote partial derivatives computed from a
scale space representation Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal th ...
L obtained by smoothing the original image with a
Gaussian kernel In mathematics, a Gaussian function, often simply referred to as a Gaussian, is a function of the base form f(x) = \exp (-x^2) and with parametric extension f(x) = a \exp\left( -\frac \right) for arbitrary real constants , and non-zero . It is ...
. In this way, the edges will be automatically obtained as continuous curves with sub-pixel accuracy. Hysteresis thresholding can also be applied to these differential and subpixel edge segments. In practice, first-order derivative approximations can be computed by central differences as described above, while second-order derivatives can be computed from the
scale space representation Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal th ...
L according to: : \begin L_(x, y) & = L(x-1, y) - 2 L(x, y) + L(x+1, y), \\ ptL_(x, y) & = \frac 1 4 (L(x-1, y-1) - L(x-1, y+1) - L(x+1, y-1) + L(x+1, y+1)), \\ ptL_(x, y) & = L(x, y-1) - 2 L(x, y) + L(x, y+1). \end corresponding to the following filter masks: : L_ = \begin 1 & -2 & 1 \end L \quad \text \quad L_ = \begin -1/4 & 0 & 1/4 \\ 0 & 0 & 0\\ 1/4 & 0 & -1/4 \end L \quad \text \quad L_ = \begin 1 \\ -2 \\ 1 \end L. Higher-order derivatives for the third-order sign condition can be obtained in an analogous fashion.


Phase congruency-based

A recent development in edge detection techniques takes a frequency domain approach to finding edge locations. Phase congruency (also known as phase coherence) methods attempt to find locations in an image where all sinusoids in the frequency domain are in phase. These locations will generally correspond to the location of a perceived edge, regardless of whether the edge is represented by a large change in intensity in the spatial domain. A key benefit of this technique is that it responds strongly to
Mach bands Mach bands is an optical illusion named after the physicist Ernst Mach. It exaggerates the contrast between edges of the slightly differing shades of gray, as soon as they contact one another, by triggering edge-detection in the human visual s ...
, and avoids false positives typically found around roof edges. A roof edge, is a discontinuity in the first order derivative of a grey-level profile.


Phase Stretch Transform (PST)

The
phase stretch transform Phase stretch transform (PST) is a computational approach to signal and image processing. One of its utilities is for feature detection and classification.M. H. Asghari, and B. Jalali, "Edge detection in digital images using dispersive ph ...
or PST is a physics-inspired computational approach to signal and image processing. One of its utilities is for feature detection and classification.M. H. Asghari, and B. Jalali
"Edge detection in digital images using dispersive phase stretch,"
International Journal of Biomedical Imaging, Vol. 2015, Article ID 687819, pp. 1–6 (2015).
PST is a spin-off from research on the
time stretch dispersive Fourier transform Time stretch dispersive Fourier transform (TS-DFT), otherwise known as time-stretch transform (TST), temporal Fourier transform or photonic time-stretch (PTS) is a spectroscopy technique that uses optical dispersion instead of a grating or prism t ...
. PST transforms the image by emulating propagation through a diffractive medium with engineered 3D dispersive property (refractive index). The operation relies on symmetry of the dispersion profile and can be understood in terms of dispersive eigenfunctions or stretch modes. PST performs similar functionality as phase contrast microscopy but on digital images. PST is also applicable to digital images as well as temporal, time series, data.


Subpixel

To increase the precision of edge detection, several subpixel techniques had been proposed, including curve-fitting, moment-based, reconstructive, and partial area effect methods. These methods have different characteristics. Curve fitting methods are computationally simple but are easily affected by noise. Moment-based methods use an integral-based approach to reduce the effect of noise, but may require more computations in some cases. Reconstructive methods use horizontal gradients or vertical gradients to build a curve and find the peak of the curve as the sub-pixel edge. Partial area effect methods are based on the hypothesis that each pixel value depends on the area at both sides of the edge inside that pixel, producing accurate individual estimation for every edge pixel. Certain variants of the moment-based technique have been shown to be the most accurate for isolated edges.


See also

* *
Edge-preserving filtering Edge-preserving smoothing or edge-preserving filtering is an image processing technique that smooths away noise or textures while retaining sharp edges. Examples are the median filter, median, bilateral filter, bilateral, guided filter, guided, anis ...
*
Feature detection (computer vision) In computer vision and image processing, a feature is a piece of information about the content of an image; typically about whether a certain region of the image has certain properties. Features may be specific structures in the image such as poi ...
for other low-level feature detectors *
Image derivative Image derivatives can be computed by using small convolution filters of size 2 × 2 or 3 × 3, such as the Discrete Laplace operator, Laplacian, Sobel operator, Sobel, Roberts cross, Roberts and Prewitt operator, Prewitt operato ...
* Gabor filter * Image noise reduction *
Kirsch operator The Kirsch operator or Kirsch compass kernel is a non-linear In mathematics and science, a nonlinear system is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engi ...
for edge detection in the compass directions *
Ridge detection In image processing, ridge detection is the attempt, via software, to locate ridges in an image, defined as curves whose points are local maxima of the function, akin to geographical ridges. For a function of ''N'' variables, its ridges are ...
for relations between edge detectors and ridge detectors *
Log Gabor filter In signal processing it is useful to simultaneously analyze the space and frequency characteristics of a signal. While the Fourier transform gives the frequency information of the signal, it is not localized. This means that we cannot determine whic ...
*
Phase stretch transform Phase stretch transform (PST) is a computational approach to signal and image processing. One of its utilities is for feature detection and classification.M. H. Asghari, and B. Jalali, "Edge detection in digital images using dispersive ph ...


References


Further reading

*
Entry on edge detection in Encyclopedia of Computer Science and EngineeringEdge Detection using FPGA
* A-contrario line segment detection with code and on-line demonstration
Edge detection using MATLAB

Subpixel edge detection using Matlab
{{DEFAULTSORT:Edge Detection Feature detection (computer vision) Image processing